Convolution in Convolution for Network in Network

نویسندگان

  • Yanwei Pang
  • Manli Sun
  • Xiaoheng Jiang
  • Xuelong Li
چکیده

Network in network (NiN) is an effective instance and an important extension of deep convolutional neural network consisting of alternating convolutional layers and pooling layers. Instead of using a linear filter for convolution, NiN utilizes shallow multilayer perceptron (MLP), a nonlinear function, to replace the linear filter. Because of the powerfulness of MLP and 1 x 1 convolutions in spatial domain, NiN has stronger ability of feature representation and hence results in better recognition performance. However, MLP itself consists of fully connected layers that give rise to a large number of parameters. In this paper, we propose to replace dense shallow MLP with sparse shallow MLP. One or more layers of the sparse shallow MLP are sparely connected in the channel dimension or channel-spatial domain. The proposed method is implemented by applying unshared convolution across the channel dimension and applying shared convolution across the spatial dimension in some computational layers. The proposed method is called convolution in convolution (CiC). The experimental results on the CIFAR10 data set, augmented CIFAR10 data set, and CIFAR100 data set demonstrate the effectiveness of the proposed CiC method.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A New Lower Bound for Completion Time Distribution Function of Stochastic PERT Networks

In this paper, a new method for developing a lower bound on exact completion time distribution function of stochastic PERT networks is provided that is based on simplifying the structure of this type of network. The designed mechanism simplifies network structure by arc duplication so that network distribution function can be calculated only with convolution and multiplication. The selection of...

متن کامل

A Deep Model for Super-resolution Enhancement from a Single Image

This study presents a method to reconstruct a high-resolution image using a deep convolution neural network. We propose a deep model, entitled Deep Block Super Resolution (DBSR), by fusing the output features of a deep convolutional network and a shallow convolutional network. In this way, our model benefits from high frequency and low frequency features extracted from deep and shallow networks...

متن کامل

A special Class of Stochastic PERT Networks

Considering the network structure is one of the new approaches in studying stochastic PERT networks (SPN). In this paper, planar networks are studied as a special class of networks. Two structural reducible mechanisms titled arc contraction and deletion are developed to convert any planar network to a series-parallel network structure.&#10In series-parallel SPN, the completion time distribution...

متن کامل

A New Lower Bound for Completion Time Distribution Function of Stochastic PERT Networks

In this paper, a new method for developing a lower bound on exact completion time distribution function of stochastic PERT networks is provided that is based on simplifying the structure of this type of network. The designed mechanism simplifies network structure by arc duplication so that network distribution function can be calculated only with convolution and multiplication. The selection of...

متن کامل

A special Class of Stochastic PERT Networks

Considering the network structure is one of the new approaches in studying stochastic PERT networks (SPN). In this paper, planar networks are studied as a special class of networks. Two structural reducible mechanisms titled arc contraction and deletion are developed to convert any planar network to a series-parallel network structure. In series-parallel SPN, the completion time distribution f...

متن کامل

Subordination and Superordination Properties for Convolution Operator

In present paper a certain convolution operator of analytic functions is defined. Moreover, subordination and superordination- preserving properties for a class of analytic operators defined on the space of normalized analytic functions in the open unit disk is obtained. We also apply this to obtain sandwich results and generalizations of some known results.

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • IEEE transactions on neural networks and learning systems

دوره   شماره 

صفحات  -

تاریخ انتشار 2017